Algorithms for distributional and adversarial pipelined filter ordering problems
نویسندگان
چکیده
منابع مشابه
Parallel Genetic Algorithms for Constrained Ordering Problems
This paper proposes two different parallel genetic algorithms (PGAs) for constrained ordering problems. Constrained ordering problems are constraint optimization problems (COPs) for which it is possible represent a candidate solution as a permutation of objects. A decoder is used to decode this permutation into an instantiafion of the COP vm-iables. Two examples of such constrmnsd ordering prob...
متن کاملDistributional Adversarial Networks
We propose a framework for adversarial training that relies on a sample rather than a single sample point as the fundamental unit of discrimination. Inspired by discrepancy measures and twosample tests between probability distributions, we propose two such distributional adversaries that operate and predict on samples, and show how they can be easily implemented on top of existing models. Vario...
متن کاملAlgorithms for Adversarial Bandit Problems with Multiple Plays
Adversarial bandit problems studied by Auer et al. [4] are multi-armed bandit problems in which no stochastic assumption is made on the nature of the process generating the rewards for actions. In this paper, we extend their theories to the case where k(≥ 1) distinct actions are selected at each time step. As algorithms to solve our problem, we analyze an extension of Exp3 [4] and an applicatio...
متن کاملCertifiable Distributional Robustness with Principled Adversarial Training
Neural networks are vulnerable to adversarial examples and researchers have proposed manyheuristic attack and defense mechanisms. We take the principled view of distributionally ro-bust optimization, which guarantees performance under adversarial input perturbations. Byconsidering a Lagrangian penalty formulation of perturbation of the underlying data distribu-tion in a Wasserst...
متن کاملDistributional Smoothing by Virtual Adversarial Examples
Smoothness regularization is a popular method to decrease generalization error. We propose a novel regularization technique that rewards local distributional smoothness (LDS), a KLdistance based measure of the model’s robustness against perturbation. The LDS is defined in terms of the direction to which the model distribution is most sensitive in the input space. We call the training with LDS r...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: ACM Transactions on Algorithms
سال: 2009
ISSN: 1549-6325,1549-6333
DOI: 10.1145/1497290.1497300